169 research outputs found

    On Integration Methods Based on Scrambled Nets of Arbitrary Size

    Full text link
    We consider the problem of evaluating I(φ):=∫[0,1)sφ(x)dxI(\varphi):=\int_{[0,1)^s}\varphi(x) dx for a function φ∈L2[0,1)s\varphi \in L^2[0,1)^{s}. In situations where I(φ)I(\varphi) can be approximated by an estimate of the form N−1∑n=0N−1φ(xn)N^{-1}\sum_{n=0}^{N-1}\varphi(x^n), with {xn}n=0N−1\{x^n\}_{n=0}^{N-1} a point set in [0,1)s[0,1)^s, it is now well known that the OP(N−1/2)O_P(N^{-1/2}) Monte Carlo convergence rate can be improved by taking for {xn}n=0N−1\{x^n\}_{n=0}^{N-1} the first N=λbmN=\lambda b^m points, λ∈{1,…,b−1}\lambda\in\{1,\dots,b-1\}, of a scrambled (t,s)(t,s)-sequence in base b≥2b\geq 2. In this paper we derive a bound for the variance of scrambled net quadrature rules which is of order o(N−1)o(N^{-1}) without any restriction on NN. As a corollary, this bound allows us to provide simple conditions to get, for any pattern of NN, an integration error of size oP(N−1/2)o_P(N^{-1/2}) for functions that depend on the quadrature size NN. Notably, we establish that sequential quasi-Monte Carlo (M. Gerber and N. Chopin, 2015, \emph{J. R. Statist. Soc. B, to appear.}) reaches the oP(N−1/2)o_P(N^{-1/2}) convergence rate for any values of NN. In a numerical study, we show that for scrambled net quadrature rules we can relax the constraint on NN without any loss of efficiency when the integrand φ\varphi is a discontinuous function while, for sequential quasi-Monte Carlo, taking N=λbmN=\lambda b^m may only provide moderate gains.Comment: 27 pages, 2 figures (final version, to appear in The Journal of Complexity

    Application of Sequential Quasi-Monte Carlo to Autonomous Positioning

    Full text link
    Sequential Monte Carlo algorithms (also known as particle filters) are popular methods to approximate filtering (and related) distributions of state-space models. However, they converge at the slow 1/N1/\sqrt{N} rate, which may be an issue in real-time data-intensive scenarios. We give a brief outline of SQMC (Sequential Quasi-Monte Carlo), a variant of SMC based on low-discrepancy point sets proposed by Gerber and Chopin (2015), which converges at a faster rate, and we illustrate the greater performance of SQMC on autonomous positioning problems.Comment: 5 pages, 4 figure

    Negative association, ordering and convergence of resampling methods

    Get PDF
    We study convergence and convergence rates for resampling schemes. Our first main result is a general consistency theorem based on the notion of negative association, which is applied to establish the almost-sure weak convergence of measures output from Kitagawa's (1996) stratified resampling method. Carpenter et al's (1999) systematic resampling method is similar in structure but can fail to converge depending on the order of the input samples. We introduce a new resampling algorithm based on a stochastic rounding technique of Srinivasan (2001), which shares some attractive properties of systematic resampling, but which exhibits negative association and therefore converges irrespective of the order of the input samples. We confirm a conjecture made by Kitagawa (1996) that ordering input samples by their states in R\mathbb{R} yields a faster rate of convergence; we establish that when particles are ordered using the Hilbert curve in Rd\mathbb{R}^d, the variance of the resampling error is O(N−(1+1/d)){\scriptscriptstyle\mathcal{O}}(N^{-(1+1/d)}) under mild conditions, where NN is the number of particles. We use these results to establish asymptotic properties of particle algorithms based on resampling schemes that differ from multinomial resampling.Comment: 54 pages, including 30 pages of supplementary materials (a typo in Algorithm 1 has been corrected

    Convergence of sequential quasi-Monte Carlo smoothing algorithms

    Get PDF
    International audienc

    Higher-order stochastic integration through cubic stratification

    Full text link
    We propose two novel unbiased estimators of the integral ∫[0,1]sf(u)du\int_{[0,1]^{s}}f(u) du for a function ff, which depend on a smoothness parameter r∈Nr\in\mathbb{N}. The first estimator integrates exactly the polynomials of degrees p<rp<r and achieves the optimal error n−1/2−r/sn^{-1/2-r/s} (where nn is the number of evaluations of ff) when ff is rr times continuously differentiable. The second estimator is computationally cheaper but it is restricted to functions that vanish on the boundary of [0,1]s[0,1]^s. The construction of the two estimators relies on a combination of cubic stratification and control ariates based on numerical derivatives. We provide numerical evidence that they show good performance even for moderate values of nn
    • …
    corecore